784 research outputs found

    Prospects for Observing an Invisibly Decaying Higgs Boson in ttˉHt\bar{t}H Production at the LHC

    Full text link
    In this thesis we study the prospects for observing the invisibly decaying Higgs boson in the associated ttˉHt\bar tH production at the LHC. The results of the Monte Carlo simulations of signal and background processes show that there is a possibility of observing the statistically significant number of signal events required for the discovery. Moreover, the analysis can be further improved to reduce the number of false reconstructions of the WW boson. The analysis of the ttˉHt \bar t H production is independent of the model in which the Higgs boson decays into the invisible channel. There are several possibilities for models where HinvisibleH\to invisible can be of interest. For this thesis, we have studied the simplest supersymmetric model, called mSUGRA. The results of the scans of the mSUGRA model parameter space show that the regions, where the branching ratio of the lightest neutral Higgs boson to the lightest neutralino pair is high, are excluded by current experimental constraints. The hbbˉh \to b \bar b channel dominates, and the possibility for discovery in this c hannel will not be suppressed by the invisible decays. This result does not disqualify invisible channel as possible signature in other models. Chapter 4 of this thesis is based on the published paper: B.P. Kersevan, \underline{M. Malawski}, E. Richter-Was: {\em Prospects for observing an invisibly decaying Higgs boson in the ttˉHt\bar t H production at the LHC}, The European Physical Journal C - Particles and Fields, 2003, vol. 29, no. 4, pp. 541 - 548.Comment: Master of Science Thesi

    Serverless Computing for Scientific Applications

    Full text link
    Serverless computing has become an important model in cloud computing and influenced the design of many applications. Here, we provide our perspective on how the recent landscape of serverless computing for scientific applications looks like. We discuss the advantages and problems with serverless computing for scientific applications, and based on the analysis of existing solutions and approaches, we propose a science-oriented architecture for a serverless computing framework that is based on the existing designs. Finally, we provide an outlook of current trends and future directions

    Flexible And Secure Access To Computing Clusters

    Get PDF
    The investigation presented in this paper was prompted by the need to provide a manageablesolution for secure access to computing clusters with a federated authentication framework.This requirement is especially important for scientists who need direct access to computingnodes in order to run their applications (e.g. chemical or medical simulations) with proprietary,open-source or custom-developed software packages. Our existing software, whichenables non-Web clients to use Shibboleth-secured services, has been extended to providedirect SSH access to cluster nodes using the Linux Pluggable Authentication Modules mechanism.This allows Shibboleth users to run the required software on clusters. Validationand performance comparison with existing SSH authentication mechanisms confirm that thepresented tools satisfy the stated requirements

    Using Unused: Non-Invasive Dynamic FaaS Infrastructure with HPC-Whisk

    Full text link
    Modern HPC workload managers and their careful tuning contribute to the high utilization of HPC clusters. However, due to inevitable uncertainty it is impossible to completely avoid node idleness. Although such idle slots are usually too short for any HPC job, they are too long to ignore them. Function-as-a-Service (FaaS) paradigm promisingly fills this gap, and can be a good match, as typical FaaS functions last seconds, not hours. Here we show how to build a FaaS infrastructure on idle nodes in an HPC cluster in such a way that it does not affect the performance of the HPC jobs significantly. We dynamically adapt to a changing set of idle physical machines, by integrating open-source software Slurm and OpenWhisk. We designed and implemented a prototype solution that allowed us to cover up to 90\% of the idle time slots on a 50k-core cluster that runs production workloads

    Virtual patients in a behavioral medicine massive open online course (MOOC) : a case-based analysis of technical capacity and user navigation pathways

    Get PDF
    BACKGROUND: Massive open online courses (MOOCs) have been criticized for focusing on presentation of short video clip lectures and asking theoretical multiple-choice questions. A potential way of vitalizing these educational activities in the health sciences is to introduce virtual patients. Experiences from such extensions in MOOCs have not previously been reported in the literature. OBJECTIVE: This study analyzes technical challenges and solutions for offering virtual patients in health-related MOOCs and describes patterns of virtual patient use in one such course. Our aims are to reduce the technical uncertainty related to these extensions, point to aspects that could be optimized for a better learner experience, and raise prospective research questions by describing indicators of virtual patient use on a massive scale. METHODS: The Behavioral Medicine MOOC was offered by Karolinska Institutet, a medical university, on the EdX platform in the autumn of 2014. Course content was enhanced by two virtual patient scenarios presented in the OpenLabyrinth system and hosted on the VPH-Share cloud infrastructure. We analyzed web server and session logs and a participant satisfaction survey. Navigation pathways were summarized using a visual analytics tool developed for the purpose of this study. RESULTS: The number of course enrollments reached 19,236. At the official closing date, 2317 participants (12.1% of total enrollment) had declared completing the first virtual patient assignment and 1640 (8.5%) participants confirmed completion of the second virtual patient assignment. Peak activity involved 359 user sessions per day. The OpenLabyrinth system, deployed on four virtual servers, coped well with the workload. Participant survey respondents (n=479) regarded the activity as a helpful exercise in the course (83.1%). Technical challenges reported involved poor or restricted access to videos in certain areas of the world and occasional problems with lost sessions. The visual analyses of user pathways display the parts of virtual patient scenarios that elicited less interest and may have been perceived as nonchallenging options. Analyzing the user navigation pathways allowed us to detect indications of both surface and deep approaches to the content material among the MOOC participants. CONCLUSIONS: This study reported on first inclusion of virtual patients in a MOOC. It adds to the body of knowledge by demonstrating how a biomedical cloud provider service can ensure technical capacity and flexible design of a virtual patient platform on a massive scale. The study also presents a new way of analyzing the use of branched virtual patients by visualization of user navigation pathways. Suggestions are offered on improvements to the design of virtual patients in MOOCs

    Never born proteins as a test case for ab initio protein structures prediction

    Get PDF
    The number of natural proteins although large is significantly smaller than the theoretical number of proteins that can be obtained combining the 20 natural amino acids, the so-called “never born proteins” (NBPs). The study of the structure and properties of these proteins allows to investigate the sources of the natural proteins being of unique characteristics or special properties. However the structural study of NPBs can also been intended as an ideal test for evaluating the efficiency of software packages for the ab initio protein structure prediction. In this research, 10.000 three-dimensional structures of proteins of completely random sequence generated according to ROSETTA and FOD model were compared. The results show the limits of these software packages, but at the same time indicate that in many cases there is a significant agreement between the prediction obtained

    A Platform for Collaborative e-Science Applications

    Get PDF
    Abstract A novel, holistic, approach to scientific investigations should, besides analysis of individual phenomena, integrate different, interdisciplinary sources of knowledge about a complex system to obtain a deep understanding of the system as a whole. This innovative way of research, recently called system-level science [1], requires advanced software environments to support collaborating research groups. Most problem-solving environments and virtual laboratories In the ViroLab project The Virtual Laboratory (see The Experiment Planning Environment supports rapid experiment plan development while the Experiment Management Interface enables loading and execution of experiments. The Experiment Repository developers and published for future use. The virtual laboratory engi Operation Invoker which instantiates grid object repr operation invocations. The GridSpace Applic load balancing on computational servers. The Data Access Service remote databases located in research institutions and Fig. 1. Architecture of the Virtual Laboratory The provenance approach in the ViroLab virtual laboratory ontology-based semantic modeling, monitoring of infrastructure, and database technologies, in order to coll the execution of experiments, represent it in a meaningful way, repository. In the ViroLab project, this virtual laboratory is used to plan and virological experiments, with various types of analysis of as the calculation of drug resistance, querying historical and about experiments, a drug resistance system based on the Retrogram been applied to other application domains, such as comparison, data mining using the Weka library, series of Gaussian application on the EGEE infrastructure. computer science classes. We have developed an environment for collaborative planning, execution of e-Science applications. It facilitates fast, close cooperation and users so it may be used by groups of experts running In-silico experiments undergo frequent changes, this platform encourages quick, agile simulation software releasing

    Search for Physics beyond the Standard Model in Events with Overlapping Photons and Jets

    Get PDF
    Results are reported from a search for new particles that decay into a photon and two gluons, in events with jets. Novel jet substructure techniques are developed that allow photons to be identified in an environment densely populated with hadrons. The analyzed proton-proton collision data were collected by the CMS experiment at the LHC, in 2016 at root s = 13 TeV, and correspond to an integrated luminosity of 35.9 fb(-1). The spectra of total transverse hadronic energy of candidate events are examined for deviations from the standard model predictions. No statistically significant excess is observed over the expected background. The first cross section limits on new physics processes resulting in such events are set. The results are interpreted as upper limits on the rate of gluino pair production, utilizing a simplified stealth supersymmetry model. The excluded gluino masses extend up to 1.7 TeV, for a neutralino mass of 200 GeV and exceed previous mass constraints set by analyses targeting events with isolated photons.Peer reviewe

    Measurement of the Jet Mass Distribution and Top Quark Mass in Hadronic Decays of Boosted Top Quarks in pp Collisions at root s=13 TeV

    Get PDF
    A measurement is reported of the jet mass distribution in hadronic decays of boosted top quarks produced in pp collisions at root s = 13 TeV. The data were collected with the CMS detector at the LHC and correspond to an integrated luminosity of 35.9 fb(-1). The measurement is performed in the lepton + jets channel of t (t) over bar events, where the lepton is an electron or muon. The products of the hadronic top quark decay t -> bW -> bq (q) over bar' are reconstructed as a single jet with transverse momentum larger than 400 GeV. The t (t) over bar cross section as a function of the jet mass is unfolded at the particle level and used to extract a value of the top quark mass of 172.6 +/- 2.5 GeV. A novel jet reconstruction technique is used for the first time at the LHC, which improves the precision by a factor of 3 relative to an earlier measurement. This highlights the potential of measurements using boosted top quarks, where the new technique will enable future precision measurements.Peer reviewe

    Measurement of the top quark forward-backward production asymmetry and the anomalous chromoelectric and chromomagnetic moments in pp collisions at √s = 13 TeV

    Get PDF
    Abstract The parton-level top quark (t) forward-backward asymmetry and the anomalous chromoelectric (d̂ t) and chromomagnetic (μ̂ t) moments have been measured using LHC pp collisions at a center-of-mass energy of 13 TeV, collected in the CMS detector in a data sample corresponding to an integrated luminosity of 35.9 fb−1. The linearized variable AFB(1) is used to approximate the asymmetry. Candidate t t ¯ events decaying to a muon or electron and jets in final states with low and high Lorentz boosts are selected and reconstructed using a fit of the kinematic distributions of the decay products to those expected for t t ¯ final states. The values found for the parameters are AFB(1)=0.048−0.087+0.095(stat)−0.029+0.020(syst),μ̂t=−0.024−0.009+0.013(stat)−0.011+0.016(syst), and a limit is placed on the magnitude of | d̂ t| < 0.03 at 95% confidence level. [Figure not available: see fulltext.
    corecore